Accessibility settings

Published on in Vol 10 (2026)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/76542, first published .
Evaluating Biomedical Feature Fusion on Machine Learning’s Predictability and Interpretability of COVID-19 Severity Types: Model Development, Interpretation, and Validation

Evaluating Biomedical Feature Fusion on Machine Learning’s Predictability and Interpretability of COVID-19 Severity Types: Model Development, Interpretation, and Validation

Evaluating Biomedical Feature Fusion on Machine Learning’s Predictability and Interpretability of COVID-19 Severity Types: Model Development, Interpretation, and Validation

1Department of Mathematics and Statistics, College of Science, University of North Carolina at Charlotte, 9201 University City Boulevard, Charlotte, NC, United States

2Department of Epidemiology and Community Health, University of North Carolina at Charlotte, Charlotte, NC, United States

Corresponding Author:

Haleigh Noelle West-Page, MS


Background: Accurately differentiating severe from nonsevere COVID-19 clinical types is critical for the health care system to optimize workflow. Current techniques lack the ability to accurately classify COVID-19 clinical types in patients, especially as SARS-CoV-2 continues to mutate.

Objective: We explore the predictability and interpretability of multiple state-of-the-art machine learning (ML) techniques trained and tested under different biomedical data types and SARS-CoV-2 variants.

Methods: Comprehensive patient-level data were collected from 362 patients (severe COVID-19: n=148; nonsevere COVID-19: n=214) infected with the original SARS-CoV-2 strain in 2020 and 1000 patients (severe COVID-19: n=500; nonsevere COVID-19: n=500) infected with the Omicron variant in 2022‐2023. The data included 26 biochemical features from blood testing and 26 clinical features from patients’ clinical characteristics and medical history. Different ML techniques, including penalized logistic regression, random forest, k-nearest neighbors, and support vector machines, were applied to build predictive classification models based on each data modality separately and together for each variant. Fifty randomized train-test splits were conducted per scenario, and performance results were recorded.

Results: The fusion (hybrid) characteristic modality yielded the highest mean area under the curve (AUC) in this study, achieving 0.915, while the biochemical and clinical modalities had AUCs of 0.862 and 0.818, respectively. All ML models performed similarly under different testing scenarios and were consistent when cross-tested with data of patients infected with the original strain and those infected with the Omicron variant. Our models ranked elevated d-dimer (biochemical), elevated high sensitivity troponin I (biochemical), and age greater than 55 years (clinical) as the most positively predictive features of severe COVID-19.

Conclusions: These results are compatible with the hypothesis that ML is a useful tool for predicting severe COVID-19 based on comprehensive individual patient–level data. Further, ML models trained on the biochemical and clinical modalities together show patterns consistent with enhanced predictive performance. The improved performance observed with Omicron variant data agrees with the hypothesis that ML approaches may retain utility across variants in this study setting, although further validation is required before clinical application. Future work using larger datasets with more ethnic variation and investigating unbiased ML interpretation methods may be able to provide further validation.

JMIR Form Res 2026;10:e76542

doi:10.2196/76542

Keywords



The COVID-19 pandemic caused by SARS-CoV-2 has impacted health care systems everywhere. Since 2019, several major SARS-CoV-2 variants and subvariants have manifested, with the Omicron variant being the most persistent since November 2021 [1]. A critical effect of the pandemic has been the sudden increased burden on health care facilities, mostly hospitals. The influx of patients with severe COVID-19 overwhelms intensive care units, which results in increased mortality [2], especially in regions with fewer health resources [3,4].

In current clinical practice, patients with COVID-19 are typically classified as having severe disease by features such as shortness of breath, low oxygen saturation, and low partial pressure of oxygen in arterial blood/fraction of inspired oxygen. However, these few features cannot sufficiently distinguish between patients with severe and those with nonsevere COVID-19, as some patients with severe COVID-19 may lack these or any symptoms upon admission [5]. Without suitable medical intervention, these patients may progress quickly to a critical condition, resulting in a high risk of mortality [6]. This uncertainty motivates a predictive method of classifying patient types, which is reliable and efficient, while also making use of alternative features. Early determination of patient types may enable health care professionals to improve their treatment plans and optimize facility resources.

The interest in integrating machine learning (ML) into general clinical practice has grown rapidly in recent years [7]. Particularly, studies on the implementation of ML as a method for clinical decision support systems (CDSSs) are ongoing. While these studies have shown great potential, their greatest limitation is a lack of interpretability [7]. Given the weight of their decisions, clinicians are hesitant to rely on “black-box” systems. In response, the subfield of explainable artificial intelligence has emerged to provide clinicians with more transparent ML models [8]. This direction of work seeks to incorporate mechanisms within ML pipelines that output both reliable classification predictions and understandable decision processes. Early detection of COVID-19 severity in patients, using ML, is often studied using a singular ML technique or data modality [9,10]. Among the studies using multiple data types or ML techniques [11], many only used data from the early waves of infection. Some attempts to provide interpretable models were found [11], but many lacked this feature.

In this study, we investigated the performance and feature importance of various ML techniques for COVID-19 severity classification prediction, and then we evaluated feature modalities that provide the most predictive and consistent results. We trained ML models using different techniques with patient-level biochemical and clinical feature modalities, both separately and together as a fusion modality. We applied logistic regression (LR), decision tree–based random forest (RF), k-nearest neighbors (kNN), and support vector machines (SVM), and we evaluated their abilities to predict severe COVID-19. We developed these ML models from data collected from patients infected with the original strain and Omicron variant to investigate model consistency across different variants within this study.


Data Collection

Our study uses two distinct datasets covering two time periods with distinct dominant viral variants. All patients were confirmed to be positive for COVID-19 by two independent quantitative reverse transcriptase–polymerase chain reaction (qRT-PCR) tests before inclusion in this study. The first set includes 362 patients infected with the original SARS-CoV-2 strain upon admission to Wuhan Union Hospital in China from January to March 2020. This dataset was previously described and analyzed by Chen et al [5] and serves as a comparative baseline in this study. Among these 362 patients, 148 had severe COVID-19 according to the guidelines established by the National Health Commission of China and the American Thoracic Society [12,13], while the remaining 214 were designated as having the nonsevere type. Patients were categorized as having severe COVID-19 based on at least one of the following criteria: (1) respiratory rate of >30 breaths per minute, (2) oxygen saturation of <93% at rest, or (3) partial pressure of oxygen in arterial blood/fraction of inspired oxygen of <300 mm Hg (40 kPa). As this dataset contains data from patients infected by the original SARS-CoV-2 strain, this set is referred to as “original” hereinafter. The second dataset consists of 1000 patients admitted to Wuhan Union Hospital in China from December 2022 to January 2023, during which time patients were diagnosed with the SARS-CoV-2 Omicron variant. Based on the same guidelines outlined earlier, 500 of these patients were classified as having severe COVID-19, while the other 500 were classified as having nonsevere COVID-19.

In our study, the input data were complete and without any missing feature data. The deidentified patient information comprised two main modalities of biomedical features. The first feature modality had 26 distinct laboratory testing features from blood tests, most of which were continuous values of the readings. The specifics of these tests are reported in detail in our prior study [5]. We refer to this feature modality as “biochemical” hereinafter. The second is a total of 26 features of one-hot encoded binary values indicating the presence of preexisting conditions, comorbidities, symptoms, and other key risk factors such as demographic information. This modality is referred to as “clinical” features hereinafter. A complete description of these features across the two modalities is presented in the supplementary materials of our prior study [5]. Together, features from both modalities were appended into a single corpus of deidentified patient data with 52 multimodal features. This was referred to as the “fusion” set, as it fused across the continuous, real-valued biochemical and binary clinical feature modalities. We note that the specific features of respiratory rate, oxygen saturation, and fraction of inspired oxygen were excluded from our predictive feature list, as they were the original clinical standard to determine COVID-19 severity.

ML Pipeline Development, Validation, and Interpretation

We developed, evaluated, and compared the performance of several state-of-the-art ML classification techniques, including RF, kNN, and SVM. All ML techniques were implemented as supervised binary classification problems. To acknowledge LR’s popularity in the field [11,14-19], we included it as a benchmark method. LR is generally sensitive to highly correlated variables, also known as multicollinearity, making predictions less precise [20]. This technique is also rather susceptible to overfitting; hence, it may be less capable of generalizing to unseen prediction sets. A powerful method of reducing overfitting is through regularization or penalization of regression. When LR is penalized (usually using l1 or l2 norms), the multicollinearity issue can be reduced [16].

To contrast with LR, we evaluated RF, kNN, and SVM. RF is an ensemble learning algorithm that adapts to nonlinearities within the data [21]. Within the “forest,” each individual binary decision tree is built by recursively partitioning subsets of the training data according to the features that yield the most information gain. Like LR, RF is interpretable after training and is a frequent choice for CDSS tasks. The kNN classifier assigns samples to their predicted classes with which they share the most similarities, as determined by a chosen distance function [22]. This technique’s performance depends on the choice of number of “neighbors” to the sample that the algorithm consults when predicting a class. We included kNN, since it is a popular classifier due to its ease of use and effectiveness when dealing with larger datasets [23]. Lastly, SVMs are classifiers trained to create boundaries between classes in the high-dimensional feature space. SVM aims to maximize the distance between samples of different classes. This decision boundary is referred to as a hyperplane, and its geometry allows for application to both linear and nonlinear problems. SVM is also a frequent choice of classification algorithm for CDSS tasks [24,25].

Using Python 3.10 and a variety of ML-related libraries from scikit-learn [26], we developed an end-to-end ML framework for each of the four classifiers to predict COVID-19 severity types from patients’ clinical, biochemical, and fusion feature modalities. A full list of packages is included in Multimedia Appendix 1. The nonsevere and severe COVID-19 types were labeled as 0 and 1, respectively. Each classifier was constructed to predict the outcome (0 or 1) based on the input features provided. We provide a graphical display of this pipeline in Figure 1.

Figure 1. Machine learning (ML) model pipeline design. Deidentified patient biomedical data were collected from Wuhan Union Hospital from January to March 2020 (patients infected with the original SARS-CoV-2 strain: n=362) and December 2022 to January 2023 (patients infected with the Omicron variant: n=1000). The data were grouped by their feature modality: biochemical, clinical, or fusion. An 80:20 split was performed to generate training and testing sets, respectively. Four classifiers were developed and evaluated for their performance: logistic regression (LR), random forest (RF), k-nearest neighbors (kNN), and support vector machine (SVM). Upon hyperparameter tuning, each classifier was validated with the same-variant hold-out set and the preprocessed cross-variant set.

For a given dataset, the corpus was first randomly partitioned into training and hold-out testing sets by an 80% to 20% split, respectively. Each feature type—biochemical, clinical and fusion—was then preprocessed by a standard scaler separately prior to training to ensure consistency across different features. Following the scaling step, a grid search method of hyperparameter tuning was used during training to maximize the ML model’s performance. The resulting optimal hyperparameters for each classifier are detailed in Multimedia Appendix 1. Upon completion, the model with trained hyperparameters was then applied to the 20% hold-out data for testing. This process was repeated 50 times to generate different random training-testing splits and hyperparameter searches, each resulting in an independent model. To summarize, each model only witnessed 80% of a given dataset for training and was evaluated for its performance on the remaining 20% of unseen data. This process was conducted to avoid overfitting [27] and establish an average performance for each setting.

ML classifiers’ performances were evaluated by the receiver operating characteristic (ROC) curve and computing the area under the curve (AUC) of ROC. In this study, AUC was selected as the main performance metric (as opposed to the F-measure or accuracy) because AUC has been shown to be more reliable than the other metrics [28]. For both the original strain and Omicron variant datasets, we evaluated ML classifier performance of training and testing on each modality separately and fused.

Each dataset (the binary clinical feature modality alone, the continuous biochemical feature modality alone, and a fusion modality that incorporates both feature modalities) underwent the pipeline defined earlier. To evaluate the consistency of the developed ML classifiers, we swapped and cross-tested with testing data from the other variant. For example, models trained on the original SARS-CoV-2 strain data were also tested with Omicron variant data, and vice versa: this process was mirrored for classifiers trained on Omicron variant data. For cross-testing, the testing data were standardized according to the scaling scheme from the classifier’s training data. During cross-testing, the entire corpus was used as a hold-out testing set, since the classifiers were only trained with one variant and never trained with the other variant’s data. The cross-set testing was evaluated for its performance exactly as the same-set testing.

One of the advantages of certain ML classification techniques is their interpretability in addition to performance. Of the classifiers developed in this study, we gathered insights from LR and RF classifiers. The resulting feature coefficient vector obtained from training LR indicated what the classifier “learned” from the data. This appears as the largest absolute coefficient corresponding to the most influencing feature in predicting the severe COVID-19 type [29]. For RF, feature importance was quantified and ranked by the feature’s mean decrease in Gini impurity [30], which is commonly used in feature selection tasks [31]. During RF’s training, these feature importances were computed using scikit-learn’s feature_importances_ package [30]. By averaging the feature rankings (ie, importance in predicting the severe COVID-19 type) over 50 runs, we compared feature importance identified by LR versus that identified by RF, as well as different feature importances between the original strain and Omicron variant data. These comprehensive investigations enabled us to validate findings from the ML classifiers by cross-checking results with other studies performing traditional statistical studies aimed at identifying predictive features of the severe COVID-19 type.

Ethical Considerations

All patients were comprehensively evaluated before being admitted to the hospitals. Their fully deidentified, anonymous biomedical data were extracted from the electronic health record system. All participants were informed about the study, agreed to participate, and provided written informed consent. An institutional review board (IRB) application was submitted and approved by the Wuhan Union Hospital, Tongji College of Medicine, Huazhong University of Science and Technology (IRB approval #IEC-J-345), where the data were collected.


ML Classifiers’ Performance

Upon running each ML classifier pipeline for 50 independent repetitions, the average AUC value was calculated (Table 1). We validated that the computed average AUC was equal to the AUC of the composite ROC curve; hence, there is no need for their distinction. These values were tabulated according to which SARS-CoV-2 variant dataset and modality were used for training each ML classifier. The SDs of the AUC did not exceed 0.06 across all testing scenarios. A summary of the SDs is provided in Multimedia Appendix 1.

Table 1. Areas under the receiver operating characteristic curve.
Modality and testing setsTraining sets
Logistic regressionRandom forestk-Nearest neighborsSupport vector machine
OriginalOmicronOriginalOmicronOriginalOmicronOriginalOmicron
Biochemical
Original0.6670.6810.6780.7390.6080.6640.6710.676
Omicron0.7460.8490.7770.8620.6900.8000.7400.853
Clinical
Original0.7200.7680.7080.7570.6680.7350.7280.769
Omicron0.7540.8180.7460.7920.7240.7890.7820.808
Fusion
Original0.7490.7540.6970.7630.6920.7330.7390.749
Omicron0.7980.9150.8090.8930.7910.8580.8270.908

Our tuned ML classifiers demonstrated overall high AUCs in predicting patients with severe COVID-19 in this study setting. In our study, ML models trained from Omicron variant data performed the best across all scenarios in our design. The highest AUC among classifiers trained on Omicron data was 0.915, compared to the original strain data’s highest AUC at 0.827 (Table 1). Performance of different ML classifiers trained on the same dataset showed minor differences. All ML classifiers developed from either original or Omicron data performed similarly when tested on the Omicron data.

For each ML classifier, ROC plots were generated for the same-variant and cross-variant testings, yielding a total of 4 training-testing combinations (original-original, Omicron-Omicron, original-Omicron, and Omicron-original, where the latter two were cross-variant testings). Each ROC plot visualizes comparisons among the three feature modalities (clinical alone, biochemical alone, and fusion). All 16 plots are presented in Multimedia Appendix 1. For brevity, we discuss four graphs from RF in Figures 2A-2D. Each graph shows the composite ROC plot over 50 independent repetitions for each modality. The shaded regions denote 1 SD from the mean of the true positive rate. The red, green, and blue lines represent the performance of models with biochemical, clinical, and fusion feature modalities, respectively.

Classifiers trained with the fusion feature modality generally demonstrated relatively higher predictive power than either biochemical or clinical feature modality alone within this study. The clinical feature modality performed slightly worse than the fusion modality, while the biochemical feature modality alone had the relative least predictive power. Regardless of ML classification technique, models trained and tested with original strain data experienced the largest variation in their performance. We note that this combination also had the lowest performance of the four training-testing combinations. Classifiers trained with original strain data had improved performance when cross-testing with Omicron data.

This table displays the mean AUC values of each ML classification technique when applied to various training and testing combinations over 50 random independent splits. For each classifier, a model was developed, trained, and tuned with either data of patients infected with the original strain or Omicron variant. The developed classifier was tested on hold-out data from both the same variant and the cross-variant data. Values are separated by training-testing combinations, as well as feature modality for training.

Figure 2. ROC curve plots for random forest (RF). This figure shows the mean ROC curves plotted for each of the four training-testing combinations. Each figure contains the ROC curve of each data modality (biochemical: red, clinical: green, and fusion: blue). Panel (A) plots the curve of classifiers trained and tested on original strain data only (ie, original-original combination). Panel (B) plots the curve of classifiers of the original-Omicron combination. Panel (C) plots the curve of classifiers of the Omicron-original combination. Panel (D) plots the curve of classifiers of the Omicron-Omicron combination. AUC: area under the curve; ROC: receiver operating characteristic.

Feature Importance Ranking

During each replication of the ML classifier, the feature coefficient vectors from the tuned LR classifier and the mean decrease of Gini impurity score vectors from the RF classifier were recorded and averaged after 50 replications. The averaged coefficients were then used to identify potential key features that differentiate severe from nonsevere COVID-19 and to evaluate how different training data (original strain or Omicron variant) and feature modalities influence the feature rankings. Note that LR’s associated coefficient vector has real values, while RF’s reduction of Gini importance is interpreted as probabilities in [0, 1].

Feature importances were determined by the fusion modality, as feature fusion demonstrated the relative highest predictive power for severe COVID-19 within this study. These feature importances are displayed in Figure 3. Regardless of ML classifiers (LR and RF) or SARS-CoV-2 variant, features such as d-dimer (biochemical modality), high sensitivity troponin I (hsTNI; biochemical), and age of >55 years (clinical modality) were consistently ranked in the top five most predictive features for COVID-19 severity. Features that often appeared among the top ten also include high-sensitivity C-reactive protein (hsCRP; biochemical) and hypertension (clinical).

There were also some disagreements in feature rankings between the two techniques. For instance, LR suggested a history of chronic obstructive pulmonary disease (COPD; clinical) as an important predictive feature of the severe COVID-19 type when trained on both the original strain and Omicron variant data, but COPD was not identified as a top feature in RF. Conversely, only RF identified elevated lymphocytes (biochemical), ferritin (biochemical), and interleukin-6 (biochemical) as important features regardless of SARS-CoV-2 variants.

By comparing the feature rankings across variants, LR trained on the original strain data identified low-, mid-, and high-grade fever (clinical) all among the top ten most predictive features, while its counterpart trained on Omicron variant data identified elevated procalcitonin (biochemical), neutrophil percentage (biochemical), and white blood cell count (biochemical) as the most predictive features. Such discrepancies in feature rankings were not observed in results from RF classifiers trained on different variants’ datasets.

Lastly, there were some slight differences in the range of feature importance (quantified by coefficients in LR and Gini impurity scores in RF) across the two variants. LR’s feature coefficients on average fell in the range of −0.95 to 2.30 for the original strain, whereas the range was −0.70 to 2.85 for the Omicron variant. Mean decreases of Gini impurity were in the range of 0-0.10 and 0-0.12 for the original strain and Omicron variant, respectively.

Figure 3. Feature rankings. This figure displays the mean feature rankings from logistic regression (LR) classifiers trained on (A) the original strain dataset and (B) on the Omicron variant, and from random forest (RF) on (C) the original strain and (D) the Omicron variant. Feature rankings are determined by the coefficients of the feature weight vector in LR, and Gini impurity scores in RF.

ML Performance and Interpretability

In this study, we evaluated the predictive power of multiple ML techniques when using different feature modalities. We found differences in model performance and interpretations across different SARS-CoV-2 variants. Overall, our results are compatible with the hypothesis that ML is a useful tool for predicting severe COVID-19 based on comprehensive individual patient–level data. More importantly, we found evidence that fusion of the biochemical and clinical modalities showed a pattern of enhanced predictive power of all ML models evaluated in this study. Models trained on multiple feature modalities have yielded the relative best performance in many metrics across all testing sets. This pattern is worthy of further investigation, as these multimodal features are accessible by health care systems, especially with wide adoption of electronic health record systems. Results can be obtained efficiently from these systems, allowing the predictive ML classification model to be a fast and reliable CDSS tool to identify patients at high risk of severe COVID-19 [32].

The similarity of performance among the four ML techniques evaluated in this study suggests that the specific choice of modeling technique is not important for the task of classifying severe from nonsevere COVID-19 types. In general, LR, RF, and SVMs all showed relatively similar performance with their highest AUC scores being 0.915, 0.893, and 0.908, respectively. The kNN model exhibits the weakest performance of the methods considered in this study, with its highest AUC being 0.858. This may be due to the relatively small size of the datasets, requiring further investigation with more samples [23]. Since model interpretability is important to the integrability of these ML models into CDSSs [7], LR and RF should be considered. The LR model offers the analyst information on which features are positively and negatively associated with the risk of severe COVID-19. However, LR is susceptible to multicollinearity between different features [20]. The RF model, on the other hand, is more resilient to the multicollinearity issue in the input data [21]. The RF model’s performance in this study is consistent that reported in other similar studies [5,11].

Upon validation, this study was among the first to use ML to propose potential critical biomedical features with the most predictive power of differentiating patients with severe COVID-19 across different dominant variant phases. The feature rankings provided by LR and RF may be important for clinical decision-making and provide insights into COVID-19 pathology with further clinical investigations. Our study postulates that elevated biomarkers such as D-dimer for coagulation, as well as hsTNI and hsCRP as indicators for cardiac damage, are positively associated with severe COVID-19 within this study. This finding is compatible with those of other studies that have shown that cardiovascular injury due to COVID-19 is highly associated with severe disease and adverse patient outcomes [33]. Other studies suggest that higher D-dimer is associated with higher risk of progressing to a severe stage [34]. Our findings also suggest that patients’ clinical information, such as being 55 years or older or having preexisting conditions such as hypertension and COPD, could increase the risk of progression to severe COVID-19. Other studies also agree that age and hypertension are major risk factors for severe COVID-19 [35-37]. The identifications fit well within the work toward constructing explainable ML pipelines as they provide clinicians with the machine’s decision-making process [7].

By identifying potential key risk factors associated with severe illness before its onset, ML may be able to give clinicians augmented views of patient information and the possibility of personalized treatment [38]. Throughout the COVID-19 pandemic, patients with COVID-19 in China were triaged based on their severity, in which patients with severe COVID-19 were treated at separate facilities compared to patients with nonsevere COVID-19 [39]. The designated hospitals for patients with severe disease were part of a coordinated emergency response to the surge of infections. It was reported that these response measures and designated facilities resulted in improved recovery rates for patients with severe disease [40]. These improved patient outcomes were made possible by accurate differentiation of patients types throughout the epidemic.

When comparing important features between patients infected by the original strain and those infected with the Omicron variant, we have identified a slight increase in the feature weight vector and Gini impurity values, which have not been reported before. This finding might suggest that COVID-19 severity could have become more predictable in more recent variants. This potentially explains the higher variability and lower performance of models trained and tested on the data of patients infected with the original SARS-CoV-2 strain. However, this claim requires significant further research with larger datasets over more detailed timelines. We also speculate that patient-level data may have higher quality in the Omicron wave than in that with the original strain. We pose a potential explanation for this difference in data quality.

As previously mentioned, our data originate from Wuhan Union Hospital in China in January to March 2020 and December 2022 to January 2023. Throughout the epidemic, government responses to disease prevention and control, medical care protocols, and national guidelines changed regularly [39]. In particular, the pathology, clinical manifestation, and diagnosis of SARS-CoV-2 evolved over this time frame [41]. According to one study detailing the timeline of such changes, clinical treatment protocols changed five times between January and March 2020 [41]. This same study reports only one change during the window of December 2022 to January 2023. The number of changes may be reflected in our data, as we found higher variability in the models trained on the data collected early during the epidemic.

Limitations and Future Work

This study has a variety of limitations. One hindrance to the generalizability of this ML framework for clinical decision support is the lack of variation in the study data samples. Due to the emergency of COVID-19, all biomedical data in this study were taken at the individual’s time of admission to one health care facility, and most patients were of Han Chinese ethnicity. The limited ethnicity coverage could result in a potential sampling bias, and the conclusions from this study might only apply to specific demographic groups. While this study builds an ML framework and demonstrates its feasibility in COVID-19 CDSSs, it is necessary to further validate the findings (eg, key influential clinical and biomedical features) with larger scale multicenter studies across different regions and different phases of the pandemic to increase patient representativeness [42]. We plan to identify additional studies and/or collect new data when possible to enhance data representation, especially across more demographic groups. Findings from these future studies could further evaluate the consistency of the ML workflow (eg, whether there are further variations in influential feature sets), and lead to new clinical studies and insights on the pathological mechanisms of COVID-19 prognosis in different demographic groups.

Furthermore, the generalizability of this study’s findings is limited by the differences in datasets. Our results revealed differences in performance between the two sets, resulting in the speculation that patient-level data may have higher quality in the Omicron wave than in that with the original strain. It is also possible that the difference in cohort sizes and prevalence of severe cases influenced the models’ performance metrics. Nevertheless, the two cohorts were obtained from the same hospital, and we had carefully designed the inclusion criteria to make the two cohorts as comparable as possible, especially with regard to key risk factors such as gender and age. Future research could improve this limitation by analyzing datasets with fewer differences in size and prevalence.

Another limitation is that we were not able to evaluate ML model predictability for other major SARS-CoV-2 variants, such as Alpha and Delta. Retrospective studies are needed to comprehensively evaluate the consistency of the developed ML models across different phases of the COVID-19 pandemic with different dominant variants and subvariants. These evaluations may be able to support or provide alternative hypotheses for the higher variation in predictability in the earlier data as it was previously discussed.

Considering the importance of interpretability in ML for clinical decision support, we acknowledge a limitation in our interpretation methods. For the RF models, we calculated feature importance using the mean decrease in Gini impurity. Gini impurity measures have been shown to be biased toward features with a high number of possible split points [43]. This can often result in continuous features to be favored over binary features when ranking their importance using an impurity measure. We acknowledge that this may have introduced bias to RF feature rankings of the fusion models, as the biochemical features were continuously valued whereas the clinical features were binary. This potential favoritism of the biochemical features over the clinical features in our fusion models limits the clinical interpretation of our feature ranking results.

Future research could overcome the abovementioned limitation by using another method of feature ranking. One potential method proposes an alternative way of constructing decision trees during the training phase [44]. In this study, statistical methods are used to preselect the most informative and unbiased features for constructing the trees. This results in more accurate decision trees and a reduction in the dimensionality of training data. Exploring these and other methods for debiasing feature rankings from RF classifiers would provide an opportunity to improve on this limitation.

There are a variety of future research directions suggested by this study. Notwithstanding improvements made on the limitations previously discussed, we acknowledge the existence of other emerging ML techniques worthy of investigation. Considering the desire for explainable ML models, techniques with an easily understandable decision processes would be of the highest interest. Additionally, given the tentative promise of LR as a classification tool for COVID-19 CDSSs, other regression techniques such as Lasso or Ridge regressions may be useful. Since Lasso and Ridge regressions are also interpretable, an analysis of the performance and feature importances of these techniques may provide even more insight.

While we found evidence supporting the hypothesis that individual ML techniques are a useful tool in predicting severe COVID-19, and investigating the power of advanced ensemble techniques may facilitate new analyses. Explorations of ensemble techniques involving LR, RF, kNN, and SVM are showing signs of predictive power for cardiovascular diseases using similar underlying datasets to this study [45,46]. These studies using ensemble techniques suggest that predictive power may be enhanced even when individual classifiers are not as robust [45,46]. Our study may find improved predictive power of severe COVID-19 by using more advanced ensemble methods beyond RF.

Further studies with a focus beyond COVID-19 may provide insights into the predictive power of ML classification techniques from individual-level data collected from patients with other respiratory illnesses. This is especially useful to health care systems inundated with patients infected with various influenza strains or the respiratory syncytial virus. Using similar ML techniques and leveraging the power of transfer learning, our developed ML pipeline can be further applied to studying other diseases with similar underlying datasets (eg, clinical and biochemical).

Another potential direction is the incorporation of more data modalities, such as patient-level medical imaging (including x-ray and computed tomographic scans) and multiomics data. Due to the higher dimensionality of imaging modalities in relation to biochemical and clinical modalities, more advanced ML techniques such as a deep convolutional neural network would need to be applied to handle such modalities.

Acknowledgments

The project described was supported by cooperative agreement (U01CK000677) from the US Centers for Disease Control and Prevention (CDC). Its contents are solely the responsibility of the authors and does not necessarily represent the official views of the CDC.

Funding

KM acknowledges the support of grants DMS-1847144 and DMS-2113676 from the US National Science Foundation. The funding source had no role in the acquisition of the data, design of the experiments, analyses of the findings, or writing of the manuscript.

Data Availability

An institutional review board (IRB) application was submitted and approved by the Wuhan Union Hospital, Tongji College of Medicine, Huazhong University of Science and Technology (IRB approval #IEC-J-345), where the data were collected. These data and the code created for this study are available by visiting the GitHub repository [47].

Authors' Contributions

Conceptualization: SC

Data curation: SC

Formal analysis: HWN-P

Funding acquisition: SC

Investigation: HWN-P

Methodology: HL, HNW-P, IO, KM, SC

Project administration: SC

Resources: SC

Software: HWN-P

Supervision: SC

Validation: HL, IO, KM

Writing–original draft: HWN-P

Writing–review and editing: HL, HWN-P, IO, KM, SC

Conflicts of Interest

None declared.

Multimedia Appendix 1

Additional tables and figures.

DOCX File, 1329 KB

  1. Khandia R, Singhal S, Alqahtani T, et al. Emergence of SARS-CoV-2 Omicron (B.1.1.529) variant, salient features, high global health concerns and strategies to counter it amid ongoing COVID-19 pandemic. Environ Res. Jun 2022;209:112816. [CrossRef] [Medline]
  2. Duggal A, Mathews KS. Impact of ICU strain on outcomes. Curr Opin Crit Care. Oct 11, 2022;28(6):667-673. [CrossRef] [Medline]
  3. Janke AT, Mei H, Rothenberg C, Becher RD, Lin Z, Venkatesh AK. Analysis of hospital resource availability and COVID-19 mortality across the United States. J Hosp Med. Apr 2021;16(4):211-214. [CrossRef] [Medline]
  4. COVID-19 epidemiological update – 16 February 2024. World Health Organization. 2024. URL: https://www.who.int/publications/m/item/covid-19-epidemiological-update-16-february-2024 [Accessed 2024-02-16]
  5. Chen Y, Ouyang L, Bao FS, et al. A multimodality machine learning approach to differentiate severe and nonsevere COVID-19: Model development and validation. J Med Internet Res. Apr 7, 2021;23(4):e23948. [CrossRef] [Medline]
  6. Wu Z, McGoogan JM. Characteristics of and important lessons from the coronavirus disease 2019 (COVID-19) outbreak in China. JAMA. Apr 7, 2020;323(13):1239. [CrossRef]
  7. Abbas Q, Jeong W, Lee SW. Explainable AI in clinical decision support systems: A meta-analysis of methods, applications, and usability challenges. Healthcare (Basel). Aug 29, 2025;13(17):2154. [CrossRef] [Medline]
  8. Mienye ID, Obaido G, Jere N, et al. A survey of explainable artificial intelligence in healthcare: Concepts, applications, and challenges. Informatics in Medicine Unlocked. 2024;51:101587. [CrossRef]
  9. Gök EC, Olgun MO. SMOTE-NC and gradient boosting imputation based random forest classifier for predicting severity level of covid-19 patients with blood samples. Neural Comput Appl. 2021;33(22):15693-15707. [CrossRef] [Medline]
  10. Luo J, Zhou L, Feng Y, Li B, Guo S. The selection of indicators from initial blood routine test results to improve the accuracy of early prediction of COVID-19 severity. PLoS ONE. 2021;16(6):e0253329. [CrossRef] [Medline]
  11. Xiong Y, Ma Y, Ruan L, et al. Comparing different machine learning techniques for predicting COVID-19 severity. Infect Dis Poverty. Feb 17, 2022;11(1):19. [CrossRef] [Medline]
  12. Novel Coronavirus Pneumonia Diagnosis and Treatment Plan (Provisional 7th Edition). 7th ed. National Health Commission of China; 2020. URL: http://www.nhc.gov.cn/yzygj/s7652m/202003/a31191442e29474b98bfed5579d5af95.shtml [Accessed 2026-04-20]
  13. Metlay JP, Waterer GW, Long AC, et al. Diagnosis and treatment of adults with community-acquired pneumonia. An official clinical practice guideline of the American Thoracic Society and Infectious Diseases Society of America. Am J Respir Crit Care Med. Oct 1, 2019;200(7):e45-e67. [CrossRef] [Medline]
  14. Shakhovska N, Yakovyna V, Chopyak V. A new hybrid ensemble machine-learning model for severity risk assessment and post-COVID prediction system. Math Biosci Eng. Apr 13, 2022;19(6):6102-6123. [CrossRef] [Medline]
  15. Cabitza F, Campagner A, Ferrari D, et al. Development, evaluation, and validation of machine learning models for COVID-19 detection based on routine blood tests. Clin Chem Lab Med. Feb 23, 2021;59(2):421-431. [CrossRef]
  16. Rymarczyk T, Kozłowski E, Kłosowski G, Niderla K. Logistic regression for machine learning in process tomography. Sensors (Basel). Aug 2, 2019;19(15):3400. [CrossRef] [Medline]
  17. Hernández-Pereira E, Fontenla-Romero O, Bolón-Canedo V, Cancela-Barizo B, Guijarro-Berdiñas B, Alonso-Betanzos A. Machine learning techniques to predict different levels of hospital care of CoVid-19. Appl Intell (Dordr). 2022;52(6):6413-6431. [CrossRef] [Medline]
  18. Jamshidi E, Asgary A, Tavakoli N, et al. Using machine learning to predict mortality for COVID-19 patients on day 0 in the ICU. Front Digit Health. 2021;3:681608. [CrossRef] [Medline]
  19. Saegerman C, Gilbert A, Donneau AF, et al. Clinical decision support tool for diagnosis of COVID-19 in hospitals. PLOS ONE. 2021;16(3):e0247773. [CrossRef] [Medline]
  20. Ranganathan P, Pramesh CS, Aggarwal R. Common pitfalls in statistical analysis: Logistic regression. Perspect Clin Res. 2017;8(3):148-151. [CrossRef] [Medline]
  21. Schonlau M, Zou RY. The random forest algorithm for statistical learning. Stata J. Mar 2020;20(1):3-29. [CrossRef]
  22. Zhang Z. Introduction to machine learning: k-nearest neighbors. Ann Transl Med. Jun 2016;4(11):218-218. [CrossRef]
  23. Bansal M, Goyal A, Choudhary A. A comparative analysis of K-Nearest Neighbor, Genetic, Support Vector Machine, Decision Tree, and Long Short Term Memory algorithms in machine learning. Decision Analytics Journal. Jun 2022;3:100071. [CrossRef]
  24. Farhadian M, Shokouhi P, Torkzaban P. A decision support system based on support vector machine for diagnosis of periodontal disease. BMC Res Notes. Jul 13, 2020;13(1):337. [CrossRef] [Medline]
  25. Medic G, Kosaner Kließ M, Atallah L, et al. Evidence-based clinical decision support systems for the prediction and detection of three disease states in critical care: A systematic literature review. F1000Res. 2019;8:1728. [CrossRef] [Medline]
  26. Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, et al. Scikit-learn: Machine learning in Python. J Mach Learn Res. 2011;(85):2825-2830. URL: https://jmlr.org/papers/volume12/pedregosa11a/pedregosa11a.pdf [Accessed 2026-04-20]
  27. Aliferis C, Simon G. Overfitting, Underfitting and General Model Overconfidence and Under-Performance Pitfalls and Best Practices in Machine Learning and AI. In: Simon GJ, Aliferis C, editors. Artificial Intelligence and Machine Learning in Health Care and Medical Sciences: Best Practices and Pitfalls. 2024:477-524. [CrossRef] [Medline]
  28. Ling CX. Using AUC and accuracy in evaluating learning algorithms. IEEE Trans Knowl Data Eng. Mar 2005;17(3):299-310. [CrossRef]
  29. Saarela M, Jauhiainen S. Comparison of feature importance measures as explanations for classification models. SN Appl Sci. Feb 2021;3(2). [CrossRef]
  30. Feature importances with a forest of trees. scikit-learn. 2024. URL: https://scikit-learn.org/1.5/auto_examples/ensemble/plot_forest_importances.html [Accessed 2026-04-20]
  31. Zhang Y, Nie B, Du J, et al. Feature selection based on neighborhood rough sets and Gini index. PeerJ Comput Sci. 2023;9:e1711. [CrossRef] [Medline]
  32. Uslu A, Stausberg J. Value of the electronic medical record for hospital care: Update from the literature. J Med Internet Res. Dec 23, 2021;23(12):e26323. [CrossRef] [Medline]
  33. Shi S, Qin M, Shen B, et al. Association of cardiac injury with mortality in hospitalized patients with COVID-19 in Wuhan, China. JAMA Cardiol. Jul 1, 2020;5(7):802-810. [CrossRef] [Medline]
  34. Yu HH, Qin C, Chen M, Wang W, Tian DS. D-dimer level is associated with the severity of COVID-19. Thromb Res. Nov 2020;195(195):219-225. [CrossRef] [Medline]
  35. Almazeedi S, Al-Youha S, Jamal MH, et al. Characteristics, risk factors and outcomes among the first consecutive 1096 patients diagnosed with COVID-19 in Kuwait. EClinicalMedicine. Jul 2020;24:100448. [CrossRef] [Medline]
  36. Suleyman G, Fadel RA, Malette KM, et al. Clinical characteristics and morbidity associated with coronavirus disease 2019 in a series of patients in Metropolitan Detroit. JAMA Netw Open. 2020;3(6):e2012270. [CrossRef] [Medline]
  37. Yadaw AS, Li YC, Bose S, Iyengar R, Bunyavanich S, Pandey G. Clinical features of COVID-19 mortality: development and validation of a clinical prediction model. Lancet Digit Health. Oct 2020;2(10):e516-e525. [CrossRef] [Medline]
  38. Jiang T, Gradus JL, Lash TL, Fox MP. Addressing measurement error in random forests using quantitative bias analysis. Am J Epidemiol. 2021;190(9):1830-1840. [CrossRef] [Medline]
  39. Johnson KB, Wei WQ, Weeraratne D, et al. Precision medicine, AI, and the future of personalized health care. Clin Transl Sci. 2021;14(1):86-93. [CrossRef] [Medline]
  40. Wu Y, Cao Z, Yang J, et al. Innovative public strategies in response to COVID-19: A review of practices from China. Health Care Sci. 2024;3(6):383-408. [CrossRef] [Medline]
  41. China’s actions to combat the new coronavirus pneumonia outbreak. Information Office of the State Council of the People’s Republic of China. URL: http://www.gov.cn/zhengce/2020-06/07/content_5517737.htm [Accessed 2023-05-12]
  42. Wu Y, Feng X, Gong M, et al. Evolution and major changes of the diagnosis and treatment protocol for COVID-19 patients in China 2020-2023. Health Care Sci. 2023;2(3):135-152. [CrossRef] [Medline]
  43. Nembrini S, König IR, Wright MN. The revival of the Gini importance? Bioinformatics. 2018;34(21):3711-3718. [CrossRef] [Medline]
  44. Nguyen TT, Huang JZ, Nguyen TT. Unbiased feature selection in learning random forests for high-dimensional data. ScientificWorldJournal. 2015;2015:471371. [CrossRef] [Medline]
  45. Zaidi SAJ, Ghafoor A, Kim J, Abbas Z, Lee SW. HeartEnsembleNet: An innovative hybrid ensemble learning approach for cardiovascular risk prediction. Healthcare (Basel). 2025;13(5):507. [CrossRef] [Medline]
  46. Fitriyani NL, Syafrudin M, Chamidah N, et al. A novel approach utilizing bagging, histogram gradient boosting, and advanced feature selection for predicting the onset of cardiovascular diseases. Mathematics. 2025;13(13):2194. [CrossRef]
  47. Hnwestpage/fusion-ML-COVID-19. GitHub. URL: https://github.com/hnwestpage/Fusion-ML-COVID-19 [Accessed 2026-04-20]


AUC: area under the curve
CDSS: clinical decision support system
COPD: chronic obstructive pulmonary disease
hsCRP: high sensitivity C-reactive protein
hsTNI: high sensitivity troponin I
IRB: Institutional Review Board
kNN: k-nearest neighbors
LR: logistic regression
ML: machine learning
qRT-PCR: reverse transcriptase–polymerase chain reaction
RF: random forest
ROC: receiver operating characteristic
SVM: support vector machine


Edited by Alessandro Rovetta; submitted 28.Apr.2025; peer-reviewed by Carmelo Militello, Oluwadotun Catherine Balogun, Seung Won Lee; final revised version received 19.Mar.2026; accepted 01.Apr.2026; published 30.Apr.2026.

Copyright

© Haleigh Noelle West-Page, Kevin McGoff, Harrison Latimer, Isaac Olufadewa, Shi Chen. Originally published in JMIR Formative Research (https://formative.jmir.org), 30.Apr.2026.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Formative Research, is properly cited. The complete bibliographic information, a link to the original publication on https://formative.jmir.org, as well as this copyright and license information must be included.